The rapid development of deep learning relies on the scale of datasets, models, and computational power. In the fields of natural language processing and computer vision, researchers have already discovered a power-law relationship between model performance and data scale. However, in the field of robotics, especially in robot manipulation, such scalable patterns have yet to be established. A research team from Tsinghua University recently published a paper exploring data scaling laws in robot imitation learning and proposed an efficient data collection strategy that collected sufficient data in just one afternoon, enabling...